Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Pooja Kedari, Pavan Rengade, Samiksha Deshmukh, Mr. Sharad Adsure, Mrs. Deepika Jaiswal
DOI Link: https://doi.org/10.22214/ijraset.2023.52568
Certificate: View Certificate
Music plays a key role in improving your well-being, as it is one of the key sources of entertainment and inspiration to keep you going. Recent studies have shown that people respond very positively to music and that music has a great effect on a person\'s brain activity. It is often better to listen your favorite music these days. This job focuses on systems that suggest songs to users based on their state of mind. This computer vision system uses components to determine a user\'s emotion from facial expressions. When an emotion is detected, the system will suggest songs with that emotion, saving users a lot of time manually selecting and playing songs. It enables the migration of computer vision techniques for such systems to automation. To achieve this goal, we use the algorithm to classify human expressions, game play, and music tracks according to their currently detected emotions. Reduces the effort and time required to manually search for songs in the list. Recognize human facial expressions by extracting facial features using Haar Cascade and CNN algorithms.
I. INTRODUCTION
Facial expressions are one of the verbal means of conveying emotions, and those feelings can be exploited through verbal exchange and Human-Machine Interface (HMI). In contemporary technologically advanced world, exceptional music players are utilized in capabilities like flip Media, speedy ahead, Flow Multicast Streams and greater. While these features meet basic person needs, you still should manually look for the current temper amongst a massive variety of songs. That is a time-ingesting undertaking that calls for a few attempt and endurance. The main purpose is to develop an smart system that can without problems apprehend facial expressions and play music tracks based on a particular expression/emotion. Emotions generally categorized into three- glad, sad, and neutral. The principle intention of this work is to expand a clever system which can easily understand facial expressions and play tracks primarily based on them. The three feelings which might be usually categorised are glad, sad and neutral. The algorithm, Haar Cascade used for the face extraction characteristic. The proposed algorithm could be very beneficial due to the fast calculation time, which improves the system overall performance. This system has been carried out in diverse fields. Human-Computer Interaction (HCI), healing methods in healthcare, and extra. Virtual tune is generally categorized as created based on attributes consisting of artist, genre, album, language, and reputation[7]. A number of the available on line tune streaming offerings use collaborative filter out-based pointers to suggest music primarily based in your choices and listening history. but, these suggestions might not fit your current temper. This approach proposes a CNN-primarily based tune recommendation technique that uses data from multimodal sentiment analysis captured by using facial actions to improve the machine's selection making on emotions detected in actual time. Machine getting to know has to end up very famous in recent years. Some styles of machine studying strategies are extra appropriate than others, depending on the character of the application and the datasets available. For a ramification of the principle forms of studying algorithms encompass supervised, unsupervised, semi-supervised and reinforcement getting to know. Neural networks (NNs) are machine getting to know and generally green techniques for extracting critical capabilities from complicated datasets and inferring functions or models that represent the ones functions[4]. The NN makes use of the schooling facts set to teach the model first. After the model is skilled, the NN may be implemented to new or previously unseen records factors to classify the records primarily based on the previously educated fashions.
II. LITERATURE REVIEW
III. METHODOLOGY
Steps of system design, training dataset, and test images are considered and the following applicable steps are used to achieve the desired result. The training set is the data with a large amount of data stored and the test set is the input given for detection purposes.
2. Pre-processing: Pre-processing is mainly performed to remove unnecessary information from the acquired image and correct some values ??so that the values ??are the same throughout. In the preprocessing phase, the image is converted from RGB to grayscale. The eyes, nose, and mouth during pretreatment are considered the regions of interest. Detected by a cascade object detector using the Haar cascade function.
3. Facial Feature Extraction: After preprocessing, the next step is feature extraction. The extracted facial features are saved as useful information during training and testing phases. You can consider the following facial features - mouth, forehead, eyes, dimples on the face and chin, eyebrows, nose, facial wrinkles. In this work, the eyes, nose, mouth, and forehead are targeted for feature extraction as the reasons for the most attractive facial expressions. It's easy to tell if you're surprised or scared when you have wrinkles on your forehead or an open mouth. But you can never show off with a human complexion. Haar feature technology is used to extract facial features.
4. Facial Recognition: For Recognition and Classification of People's Facial Expressions Convolutional Neural Network classifier is used. Gets the best match to the test data from the training data set, so it better matches the currently known formula. Face recognition is an important computer vision problem for recognition and localization images. Face recognition can be performed with a traditional feature-based cascade classifier using the OpenCV library. Face detection can be achieved with a multitasking cascaded CNN.
5. Play Music: The last and most important part of this system is to play music based on the recognized person's current emotions. Once the user's facial expressions are classified, the corresponding user's emotional state is identified, and the collected songs of different genres by the number of emotions and put them in a list. Each emotion category has a certain number of traces. If the user expression is classified using the CNN algorithm, songs belonging to that category will be played.
A. Input
???????B. Algorithm
Convolutional Neural Networks (CNN): Convolutional Neural Networks, a type of deep learning algorithm, are very good at reading pictures. Great algorithm for automatic processing of pictures. The picture contains RGB combination information. You can use matplotlib to import an image from a file into memory. A convolutional neural network is a special type of neural community that helps machines learn and classify images.
A convolutional neural network has three types of layers
???????
The proposed work represents a facial expression recognition system to play the song according to expression detected. It uses a CNN approach for feature extraction. Our system is used to recognize the user\'s emotions based on the facial expression using Haar Feature. We integrate python code in to the web service and play music on the basis facial expressions like happy, sad or neutral. Recognizing emotions using facial expressions is one of the important research topics which has collected much attention in the past. It can be seen that emotion recognition with the help of image processing algorithms is increasing every day. Researchers are constantly working on the ways to solve it using different kinds image processing properties and methods. Henceforth, it is possible to extend the system on a mobile device.
[1] Vincenzo Moscato, Antonio Picariello And Giancarlo Sperl, \"An Emotional Recommender System For Music\", Ieee Intelligent System, October 01,2020. [2] Ahlam Alrihaili, Alaa Alsaedi, Kholood Albalawi,Liyakathunisa Syed, \"Music Recommender System For Users Based On Emotion Detection Through Facial Features\", Ieee,2020. [3] S Metilda Florence1 And M Uma2, \"Emotional Detection And Music Recommendation System Based On User Facial Expression\",Iop Conference Series: Materials Science And Engineering,2020. [4] Sushmita G. Kamble And Asso. Prof. A. H. Kulkarni, \"Facial Expression Based Music Player\", Intl. Conference On Advances In Computing, Communications And Informatics (Icacci), Sept. 21 24, 2016. [5] B. Nareen Sai1, D. Sai. Vamshi2, Piyush Pogakwar3, V. Seetharama Rao2, Y. Srinivasulu3 1, 2, 3students, 4asst. Professor, 5assc., \"Music Recommendation System Using Facial Expression Recognition Using Machine Learning\",International Journal For Research In Applied Science & Engineering Technology (Ijraset), 2022. [6] S Metilda Florence1 And M Uma2, \"Emotional Detection And Music Recommendation System Based On User Facial Expression\",Iop Conference Series: Materials Science And Engineering,2020. [7] Ziyang Yu1, Mengda Zhao1, Yilin Wu1, Peizhuo Liu1, Hexu Chen,Research On Automatic Music Recommendation Algorithm Based On \"Facial Micro-Expression Recognition\",Proceedings Of The 39th Chinese Control Conference July 27-29, 2020. [8] Dr. Sunil Bhutada, Ch. Sadhvika, Gutta.Abigna, P.Srinivas Reddy, \"Emotion Based Music Recommendation System\",Jetir April 2020. [9] Mikhail Rumiantcev, Oleksiy Khriyenko, \"Emotion Based Music Recommendation System\",_Proceeding Of The 26th Conference Of Fruct Association. [10] Krupa K S,Kartikey Rai,Ambara G,Sahil Choudhury, \"Emotion Aware Smart Music Recommender System Using Two Level Cnn\",Third International Conference On Smart Systems And Inventive Technology (Icssit 2020). [11] Karthik Subramanian Nathan?, Manasi Arun†And Megala S Kannan,Emosic- \"An Emotion Based Music Player For Android\",2017 Ieee International Symposium On Signal Processing And Information Technology (Isspit)
Copyright © 2023 Pooja Kedari, Pavan Rengade, Samiksha Deshmukh, Mr. Sharad Adsure, Mrs. Deepika Jaiswal. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET52568
Publish Date : 2023-05-19
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here